Improved Architectures and Training Algorithms for Deep Operator Networks

نویسندگان

چکیده

Operator learning techniques have recently emerged as a powerful tool for maps between infinite-dimensional Banach spaces. Trained under appropriate constraints, they can also be effective in the solution operator of partial differential equations (PDEs) an entirely self-supervised manner. In this work we analyze training dynamics deep networks (DeepONets) through lens Neural Tangent Kernel theory, and reveal bias that favors approximation functions with larger magnitudes. To correct propose to adaptively re-weight importance each example, demonstrate how procedure effectively balance magnitude back-propagated gradients during via gradient descent. We novel network architecture is more resilient vanishing pathologies. Taken together, our developments provide new insights into DeepONets consistently improve their predictive accuracy by factor 10-50x, demonstrated challenging setting PDE operators absence paired input-output observations. All code data accompanying manuscript will made publicly available at https://github.com/PredictiveIntelligenceLab/ImprovedDeepONets .

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Implementation of Evolutionary Algorithms for Deep Architectures

Deep learning is becoming an increasingly interesting and powerful machine learning method with successful applications in many domains, such as natural language processing, image recognition, and hand-written character recognition. Despite of its eminent success, limitations of traditional learning approach may still prevent deep learning from achieving a wide range of realistic learning tasks...

متن کامل

DeepArchitect: Automatically Designing and Training Deep Architectures

In deep learning, performance is strongly affected by the choice of architecture and hyperparameters. While there has been extensive work on automatic hyperparameter optimization for simple spaces, complex spaces such as the space of deep architectures remain largely unexplored. As a result, the choice of architecture is done manually by the human expert through a slow trial and error process g...

متن کامل

Deep Neuroevolution: Genetic Algorithms Are a Competitive Alternative for Training Deep Neural Networks for Reinforcement Learning

Deep artificial neural networks (DNNs) are typically trained via gradient-based learning algorithms, namely backpropagation. Evolution strategies (ES) can rival backprop-based algorithms such as Q-learning and policy gradients on challenging deep reinforcement learning (RL) problems. However, ES can be considered a gradient-based algorithm because it performs stochastic gradient descent via an ...

متن کامل

DeepDefense: Training Deep Neural Networks with Improved Robustness

Despite the efficacy on a variety of computer vision tasks, deep neural networks (DNNs) are vulnerable to adversarial attacks, limiting their applications in security-critical systems. Recent works have shown the possibility of generating imperceptibly perturbed image inputs (a.k.a., adversarial examples) to fool well-trained DNN models into making arbitrary predictions. To address this problem...

متن کامل

Stable Architectures for Deep Neural Networks

Deep neural networks have become valuable tools for supervised machine learning, e.g., in the classification of text or images. While offering superior results over traditional techniques to find and express complicated patterns in data, deep architectures are known to be challenging to design and train such that they generalize well to new data. An important issue that must be overcome is nume...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of Scientific Computing

سال: 2022

ISSN: ['1573-7691', '0885-7474']

DOI: https://doi.org/10.1007/s10915-022-01881-0